Estimating Mutual Information by Local Gaussian Approximation
نویسندگان
چکیده
Estimating Mutual Information by Local Gaussian Approximation Report Title Estimating mutual information (MI) from samples is a fundamental problem in statistics, machine learning, and data analysis. Recently it was shown that a popular class of non-parametric MI estimators perform very poorly for strongly dependent variables and have sample complexity that scales exponentially with the true MI. This undesired behavior was attributed to the reliance of those estimators on local uniformity of the underlying (and unknown) probability density function. Here we present a novel semi-parametric estimator of mutual information, where at each sample point, densities are {\em locally} approximated by a Gaussians distribution. We demonstrate that the estimator is asymptotically unbiased. We also show that the proposed estimator has a superior performance compared to several baselines, and is able to accurately measure relationship strengths over many orders of magnitude. Conference Name: The 31st Conference on Uncertainty in Artificial Intelligence Conference Date: July 13, 2015 Estimating Mutual Information by Local Gaussian Approximation Shuyang Gao Information Sciences Institute University of Southern California [email protected] Greg Ver Steeg Information Sciences Institute University of Southern California [email protected] Aram Galstyan Information Sciences Institute University of Southern California [email protected]
منابع مشابه
Measures of observation impact in non-Gaussian data assimilation
Non-Gaussian/non-linear data assimilation is becoming an increasingly important area of research in the Geosciences as the resolution and non-linearity of models are increased and more and more non-linear observation operators are being used. In this study, we look at the effect of relaxing the assumption of a Gaussian prior on the impact of observations within the data assimilation system. Thr...
متن کاملOn the Estimation of Joint Mutual Information for Physical Layer Security
In a two node communication environment, the number of key bits (IK) per observation of the reciprocal channel can be computed using the mutual information between the forward and reverse estimated channels. The mutual information can not be computed in closed form, if the probability density function (pdf) of the channel is unknown. This work studies the computation of IK for an arbitrary chan...
متن کاملNear-Optimal Sensor Placements in Gaussian Processes: Theory, Efficient Algorithms and Empirical Studies
When monitoring spatial phenomena, which can often be modeled as Gaussian processes (GPs), choosing sensor locations is a fundamental task. There are several common strategies to address this task, for example, geometry or disk models, placing sensors at the points of highest entropy (variance) in the GP model, and A-, D-, or E-optimal design. In this paper, we tackle the combinatorial optimiza...
متن کاملRAPPORT System identification with information theoretic criteria
Attention is focused in this paper on the approximation problem of system identii-cation with information theoretic criteria. For a class of problems it is shown that the criterion of mutual information rate is identical to the criterion of exponential-of-quadratic cost and to H 1 entropy. In addition the relation between the likelihood function and divergence is explored. As a consequence of t...
متن کاملAlmost Perfect Privacy for Additive Gaussian Privacy Filters
We study the maximal mutual information about a random variable Y (representing non-private information) displayed through an additive Gaussian channel when guaranteeing that only ε bits of information is leaked about a random variable X (representing private information) that is correlated with Y . Denoting this quantity by gε(X,Y ), we show that for perfect privacy, i.e., ε = 0, one has g0(X,...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2015